Skip to content

Conversation

@Pouyanpi
Copy link
Collaborator

restore skipped test

@codecov-commenter
Copy link

Codecov Report

✅ All modified and coverable lines are covered by tests.

📢 Thoughts on this report? Let us know!

@Pouyanpi Pouyanpi merged commit aa6a08e into develop Oct 28, 2025
7 checks passed
@Pouyanpi Pouyanpi deleted the test/restore-passthrough-v2 branch October 28, 2025 12:59
tgasser-nv added a commit that referenced this pull request Oct 28, 2025
* Initial checkin

* Add nemoguardrails/server to pyright type-checking

* chore(types): Type-clean embeddings/ (25 errors) (#1383)

* test: restore test that was skipped due to Colang 2.0 serialization issue (#1449)

* fix(llm): add fallback extraction for reasoning traces from <think> tags (#1474)

Adds a compatibility layer for LLM providers that don't properly populate reasoning_content in additional_kwargs. When reasoning_content is missing, the system now falls back to extracting reasoning traces from <think>...</think> tags in the response content and removes the tags from the final output.

This fixes compatibility with certain NVIDIA models (e.g., nvidia/llama-3.3-nemotron-super-49b-v1.5) in langchain-nvidia-ai-endpoints that include reasoning traces in <think> tags but fail to populate the reasoning_content field.

All reasoning models using ChatNVIDIA should expose reasoning content consistently through the same interface

* Clean up the config_id logic based on Traian and Greptile feedback

---------

Co-authored-by: Pouyan <[email protected]>
tgasser-nv added a commit that referenced this pull request Oct 28, 2025
* Initial checkin

* Add nemoguardrails/server to pyright type-checking

* chore(types): Type-clean embeddings/ (25 errors) (#1383)

* test: restore test that was skipped due to Colang 2.0 serialization issue (#1449)

* fix(llm): add fallback extraction for reasoning traces from <think> tags (#1474)

Adds a compatibility layer for LLM providers that don't properly populate reasoning_content in additional_kwargs. When reasoning_content is missing, the system now falls back to extracting reasoning traces from <think>...</think> tags in the response content and removes the tags from the final output.

This fixes compatibility with certain NVIDIA models (e.g., nvidia/llama-3.3-nemotron-super-49b-v1.5) in langchain-nvidia-ai-endpoints that include reasoning traces in <think> tags but fail to populate the reasoning_content field.

All reasoning models using ChatNVIDIA should expose reasoning content consistently through the same interface

* Clean up the config_id logic based on Traian and Greptile feedback

---------

Co-authored-by: Pouyan <[email protected]>
Pouyanpi added a commit that referenced this pull request Nov 3, 2025
* Initial checkin

* Add nemoguardrails/server to pyright type-checking

* chore(types): Type-clean embeddings/ (25 errors) (#1383)

* test: restore test that was skipped due to Colang 2.0 serialization issue (#1449)

* fix(llm): add fallback extraction for reasoning traces from <think> tags (#1474)

Adds a compatibility layer for LLM providers that don't properly populate reasoning_content in additional_kwargs. When reasoning_content is missing, the system now falls back to extracting reasoning traces from <think>...</think> tags in the response content and removes the tags from the final output.

This fixes compatibility with certain NVIDIA models (e.g., nvidia/llama-3.3-nemotron-super-49b-v1.5) in langchain-nvidia-ai-endpoints that include reasoning traces in <think> tags but fail to populate the reasoning_content field.

All reasoning models using ChatNVIDIA should expose reasoning content consistently through the same interface

* Clean up the config_id logic based on Traian and Greptile feedback

---------

Co-authored-by: Pouyan <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants